Skip to content

Conversation

@cmunley1
Copy link
Contributor

No description provided.

Signed-off-by: Christian Munley <[email protected]>
@cmunley1 cmunley1 requested review from cwing-nvidia, hwolff99 and lbliii and removed request for lbliii January 28, 2026 06:27

## Rollout Collection

### Start vLLM Server
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can we follow the pattern where we are using a hosted model to generate rollouts like the quickstart?

echo "policy_base_url: https://api.openai.com/v1
policy_api_key: your-openai-api-key
policy_model_name: gpt-4.1-2025-04-14" > env.yaml

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Unnecessary burden to get a model and serve it with vLLM, right?


## Example Usage

### GSM8K Environment
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Then unlike my comment here on Reasoning gym https://github.com/NVIDIA-NeMo/Gym/pull/617/changes#r2800137030 we do not have the "setup steps" before running ng_run

Need one pattern and to follow it


---

## Start Model Server
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Similar comment to the other env tutorials about a hosted model raising barrier to entry and consistency with quickstart.

Also this one doesn't have the instruction to pull weights from HF

## Start Model Server

```bash
uv add vllm
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@cmunley1 cmunley1 closed this Feb 12, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants